1. Introduce the Main Idea of Transparency and Interaction in Machine Learning
Transparency
Transparency aims to understand what did the model learn, what do the weights tell us about features and their importance, when an object is classified by this model, which feature values contributed to each class and how much.
Interaction
If we want to label more objects, which one should we label next? Can we provide any rationales into the learning process?
Why Transparency?
A perfect example comes from my professor Mustafa. He said that sometimes the machine instead provides a suggestion and gives a reason than deciding for the human. In medical recommender system, the doctor prefers the system provides suspected symptoms to the percentage of the diagnosis. If a cancer detection system only gives “This patient has a 79% chance gets cancer”. Such a result will make confuse the doctor and the patient. In contrast, if this system said: “This patient has a 79% chance gets cancer because I detected the patient has high compression in the spinal cord and pleural effusion”. This diagonal report will be of great help to doctors and dramatically reduce diagnostic time. Moreover, transparency also used in Trustworthy fraud detection{target:blank}. We should tell the customer this product evaluation is fake and why at the same time.
2. Concept Learning
Motivation
Different classification task has different purposes such as emails labeled as spam/~spam, patient records labeled as flu/~flu or reviews labeled as positive/negative.
Our goal is inducing a general function that fits the training data well and generalized well to unseen or future data.
General-to-Specific Ordering of Hypothese
Example
Suppose we have two features: Weight: (Light, Heavy) and Color: (Red, Green, Blue. The class is YES and NO. Therefore, we have six objects/instances . and we have 64 hypotheses/functions .
Weight | Color | F1 | … | F64 |
---|---|---|---|---|
Light | Red | NO | YES | |
Light | Green | NO | YES | |
Light | Blue | NO | YES | |
Heavy | Red | NO | YES | |
Heavy | Green | NO | YES | |
Heavy | Blue | NO | YES |
If we know that <Light, Red> is ‘Yes’, can you guess how many hypothese left? The answer is .
The number of hypotheses of <Light, Blue > is ‘Yes,’ or ‘No’ is 16.
For now, we know how to calculate the hypothesis. Let us assume two particular assumptions:
- ‘No’ to everything
- ‘Yes’ to everything
- In particular:
- : ‘No’ to everything
- : ‘Yes’ to everything
- <Light, Red>: ‘Yes’ to <Light, Red> and ‘No’ to everthing else
- <Light, >: ‘Yes’ to anything that is Light; ignore Color; everything else is ‘No’
- Combination of one or more of with other feature values is still equivalen to ‘No’ to everything. E.g., Red is ‘No’ to everything
- Most-general hypothesis is where :
- Most-specific hypothesis is where :
- These called hypothesis space
But, how do we search this space efficiently?
Algorithms
1.Find-S
Initialize to the most specific hypothesis in For each positive training instance
- For each attribute constraint in :
- If the constraint in is satisfied by
- Then do nothing
- Else replace in by the next more general constraint that is satisfied by Output hypothesis
- Warning!!!: In that case, Find-S always return the most specific one, but we do not know whether we need the most specific one or the most general one or interval between them. Moreover, what if our training hata has errors or what if the most specific hypothesis in not unique?
2.List-Then-Eliminate
VS a list of containing every hypothesis in For each training example,
- Remove from VS any hypothesis for which Output VS
Though this algorithm can output all consistent hypotheses, we need to list all possible explanation. It is possible in small predictions space. However, it is impossible for infinite hypothesis spaces.
3.Candidate-Elimination
Initialize to the set of maximally-general hypotheses in Initialize to the set of maximally-specific hypotheses in For each training example , do
- If is a positive example
- Remove from any hypothesis that is inconsistent with
- For each hypothesis in that is not consistent with
- Remove from
- Add to to all minimal gegneralizations of such that
- is consistent with , and some member of is more general than
- Remove from any hypothesis that is more general than another hypothesis in